首页> 外文OA文献 >Real-Time Facial Segmentation and Performance Capture from RGB Input
【2h】

Real-Time Facial Segmentation and Performance Capture from RGB Input

机译:RGB输入的实时面部分割和性能捕获

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

We introduce the concept of unconstrained real-time 3D facial performancecapture through explicit semantic segmentation in the RGB input. To ensurerobustness, cutting edge supervised learning approaches rely on large trainingdatasets of face images captured in the wild. While impressive tracking qualityhas been demonstrated for faces that are largely visible, any occlusion due tohair, accessories, or hand-to-face gestures would result in significant visualartifacts and loss of tracking accuracy. The modeling of occlusions has beenmostly avoided due to its immense space of appearance variability. To addressthis curse of high dimensionality, we perform tracking in unconstrained imagesassuming non-face regions can be fully masked out. Along with recentbreakthroughs in deep learning, we demonstrate that pixel-level facialsegmentation is possible in real-time by repurposing convolutional neuralnetworks designed originally for general semantic segmentation. We develop anefficient architecture based on a two-stream deconvolution network withcomplementary characteristics, and introduce carefully designed trainingsamples and data augmentation strategies for improved segmentation accuracy androbustness. We adopt a state-of-the-art regression-based facial trackingframework with segmented face images as training, and demonstrate accurate anduninterrupted facial performance capture in the presence of extreme occlusionand even side views. Furthermore, the resulting segmentation can be directlyused to composite partial 3D face models on the input images and enableseamless facial manipulation tasks, such as virtual make-up or facereplacement.
机译:我们通过在RGB输入中进行显式语义分割,引入了不受约束的实时3D面部表情捕获的概念。为了确保稳健性,最先进的监督学习方法依赖于在野外捕获的大型人脸图像训练数据集。虽然已经证明了在很大程度上可见的面部具有令人印象深刻的跟踪质量,但是由于头发,配件或手势的原因而造成的任何遮挡都会导致明显的视觉伪像并降低跟踪精度。由于其巨大的外观可变性空间,几乎避免了对遮挡的建模。为了解决这种高维诅咒,我们假设可以完全掩盖非脸部区域,因此可以在不受约束的图像中进行跟踪。伴随着深度学习领域的最新突破,我们证明了通过重新利用最初为一般语义分割而设计的卷积神经网络,可以实时实现像素级面部分割。我们基于具有互补特性的两流反卷积网络开发了一种有效的架构,并介绍了经过精心设计的训练样本和数据扩充策略,以提高分割的准确性和鲁棒性。我们采用最先进的基于回归的面部跟踪框架,并以分割的面部图像作为训练,并在极端遮挡甚至侧视图的情况下演示准确,不间断的面部表现捕获。此外,所得到的分割可以直接用于输入图像上的合成部分3D面部模型,并且可以实现无缝的面部操作任务,例如虚拟化妆或面部置换。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号